19 research outputs found
Generative Pre-trained Transformer for Vietnamese Community-based COVID-19 Question Answering
Recent studies have provided empirical evidence of the wide-ranging potential
of Generative Pre-trained Transformer (GPT), a pretrained language model, in
the field of natural language processing. GPT has been effectively employed as
a decoder within state-of-the-art (SOTA) question answering systems, yielding
exceptional performance across various tasks. However, the current research
landscape concerning GPT's application in Vietnamese remains limited. This
paper aims to address this gap by presenting an implementation of GPT-2 for
community-based question answering specifically focused on COVID-19 related
queries in Vietnamese. We introduce a novel approach by conducting a
comparative analysis of different Transformers vs SOTA models in the
community-based COVID-19 question answering dataset. The experimental findings
demonstrate that the GPT-2 models exhibit highly promising outcomes,
outperforming other SOTA models as well as previous community-based COVID-19
question answering models developed for Vietnamese
BARTPhoBEiT: Pre-trained Sequence-to-Sequence and Image Transformers Models for Vietnamese Visual Question Answering
Visual Question Answering (VQA) is an intricate and demanding task that
integrates natural language processing (NLP) and computer vision (CV),
capturing the interest of researchers. The English language, renowned for its
wealth of resources, has witnessed notable advancements in both datasets and
models designed for VQA. However, there is a lack of models that target
specific countries such as Vietnam. To address this limitation, we introduce a
transformer-based Vietnamese model named BARTPhoBEiT. This model includes
pre-trained Sequence-to-Sequence and bidirectional encoder representation from
Image Transformers in Vietnamese and evaluates Vietnamese VQA datasets.
Experimental results demonstrate that our proposed model outperforms the strong
baseline and improves the state-of-the-art in six metrics: Accuracy, Precision,
Recall, F1-score, WUPS 0.0, and WUPS 0.9
ViCLEVR: A Visual Reasoning Dataset and Hybrid Multimodal Fusion Model for Visual Question Answering in Vietnamese
In recent years, Visual Question Answering (VQA) has gained significant
attention for its diverse applications, including intelligent car assistance,
aiding visually impaired individuals, and document image information retrieval
using natural language queries. VQA requires effective integration of
information from questions and images to generate accurate answers. Neural
models for VQA have made remarkable progress on large-scale datasets, with a
primary focus on resource-rich languages like English. To address this, we
introduce the ViCLEVR dataset, a pioneering collection for evaluating various
visual reasoning capabilities in Vietnamese while mitigating biases. The
dataset comprises over 26,000 images and 30,000 question-answer pairs (QAs),
each question annotated to specify the type of reasoning involved. Leveraging
this dataset, we conduct a comprehensive analysis of contemporary visual
reasoning systems, offering valuable insights into their strengths and
limitations. Furthermore, we present PhoVIT, a comprehensive multimodal fusion
that identifies objects in images based on questions. The architecture
effectively employs transformers to enable simultaneous reasoning over textual
and visual data, merging both modalities at an early model stage. The
experimental findings demonstrate that our proposed model achieves
state-of-the-art performance across four evaluation metrics. The accompanying
code and dataset have been made publicly accessible at
\url{https://github.com/kvt0012/ViCLEVR}. This provision seeks to stimulate
advancements within the research community, fostering the development of more
multimodal fusion algorithms, specifically tailored to address the nuances of
low-resource languages, exemplified by Vietnamese.Comment: A pre-print version and submitted to journa
Facile Template In-Situ Fabrication of ZnCo2O4 Nanoparticles with Highly Photocatalytic Activities under Visible-Light Irradiation
High specific surface area ZnCo2O4 nanoparticles were prepared via a sacrificial template accelerated hydrolysis by using nanoparticles of ZnO with highly polar properties as a template. The obtained ZnCo2O4 nanoparticles were characterized by the method of scanning electron microscopy (SEM), X-ray diffraction (XRD), Brunauer-Emmett-Teller (BET) surface area measurements, Transmission electron microscopy (TEM), Raman spectroscopy, and X-ray photoelectron spectroscopy (XPS). The obtained nanoparticles were performed as a photocatalyst for the degradation of methylene blue in aqueous solution under visible irradiation. The photocatalytic degradation rate of methylene blue onto the synthesized ZnCo2O4 was higher than that of commercial ZnO and synthesized ZnO template. Copyright © 2019 BCREC Group. All rights reserved
TextANIMAR: Text-based 3D Animal Fine-Grained Retrieval
3D object retrieval is an important yet challenging task, which has drawn
more and more attention in recent years. While existing approaches have made
strides in addressing this issue, they are often limited to restricted settings
such as image and sketch queries, which are often unfriendly interactions for
common users. In order to overcome these limitations, this paper presents a
novel SHREC challenge track focusing on text-based fine-grained retrieval of 3D
animal models. Unlike previous SHREC challenge tracks, the proposed task is
considerably more challenging, requiring participants to develop innovative
approaches to tackle the problem of text-based retrieval. Despite the increased
difficulty, we believe that this task has the potential to drive useful
applications in practice and facilitate more intuitive interactions with 3D
objects. Five groups participated in our competition, submitting a total of 114
runs. While the results obtained in our competition are satisfactory, we note
that the challenges presented by this task are far from being fully solved. As
such, we provide insights into potential areas for future research and
improvements. We believe that we can help push the boundaries of 3D object
retrieval and facilitate more user-friendly interactions via vision-language
technologies.Comment: arXiv admin note: text overlap with arXiv:2304.0573
Ventilator-associated respiratory infection in a resource-restricted setting: impact and etiology.
BACKGROUND: Ventilator-associated respiratory infection (VARI) is a significant problem in resource-restricted intensive care units (ICUs), but differences in casemix and etiology means VARI in resource-restricted ICUs may be different from that found in resource-rich units. Data from these settings are vital to plan preventative interventions and assess their cost-effectiveness, but few are available. METHODS: We conducted a prospective observational study in four Vietnamese ICUs to assess the incidence and impact of VARI. Patients ≥ 16 years old and expected to be mechanically ventilated > 48 h were enrolled in the study and followed daily for 28 days following ICU admission. RESULTS: Four hundred fifty eligible patients were enrolled over 24 months, and after exclusions, 374 patients' data were analyzed. A total of 92/374 cases of VARI (21.7/1000 ventilator days) were diagnosed; 37 (9.9%) of these met ventilator-associated pneumonia (VAP) criteria (8.7/1000 ventilator days). Patients with any VARI, VAP, or VARI without VAP experienced increased hospital and ICU stay, ICU cost, and antibiotic use (p < 0.01 for all). This was also true for all VARI (p < 0.01 for all) with/without tetanus. There was no increased risk of in-hospital death in patients with VARI compared to those without (VAP HR 1.58, 95% CI 0.75-3.33, p = 0.23; VARI without VAP HR 0.40, 95% CI 0.14-1.17, p = 0.09). In patients with positive endotracheal aspirate cultures, most VARI was caused by Gram-negative organisms; the most frequent were Acinetobacter baumannii (32/73, 43.8%) Klebsiella pneumoniae (26/73, 35.6%), and Pseudomonas aeruginosa (24/73, 32.9%). 40/68 (58.8%) patients with positive cultures for these had carbapenem-resistant isolates. Patients with carbapenem-resistant VARI had significantly greater ICU costs than patients with carbapenem-susceptible isolates (6053 USD (IQR 3806-7824) vs 3131 USD (IQR 2108-7551), p = 0.04) and after correction for adequacy of initial antibiotics and APACHE II score, showed a trend towards increased risk of in-hospital death (HR 2.82, 95% CI 0.75-6.75, p = 0.15). CONCLUSIONS: VARI in a resource-restricted setting has limited impact on mortality, but shows significant association with increased patient costs, length of stay, and antibiotic use, particularly when caused by carbapenem-resistant bacteria. Evidence-based interventions to reduce VARI in these settings are urgently needed
UIT-HSE at WNUT-2020 Task 2: Exploiting CT-BERT for Identifying COVID-19 Information on the Twitter Social Network
Recently, COVID-19 has affected a variety of real-life aspects of the world and led to dreadful consequences. More and more tweets about COVID-19 has been shared publicly on Twitter. However, the plurality of those Tweets are uninformative, which is challenging to build automatic systems to detect the informative ones for useful AI applications. In this paper, we present our results at the W-NUT 2020 Shared Task 2: Identification of Informative COVID-19 English Tweets. In particular, we propose our simple but effective approach using the transformer-based models based on COVID-Twitter-BERT (CT-BERT) with different fine-tuning techniques. As a result, we achieve the F1-Score of 90.94\% with the third place on the leaderboard of this task which attracted 56 submitted teams in total